# Mathematical Reasoning Optimization

Phi 4 Reasoning Plus GGUF
MIT
Phi-4-reasoning-plus is an open-source reasoning model developed by Microsoft Research, focusing on advanced reasoning capabilities in mathematics, science, and programming fields.
Large Language Model Supports Multiple Languages
P
unsloth
109.62k
47
Microsoft Phi 4 Reasoning GGUF
MIT
This is a quantized version of Microsoft's Phi-4-reasoning model, optimized using llama.cpp for inference tasks and supporting multiple quantization options.
Large Language Model
M
bartowski
5,443
4
Mimo 7B SFT
MIT
MiMo-7B-RL is a reinforcement learning model trained based on the MiMo-7B-SFT model, achieving performance comparable to OpenAI o1-mini in mathematical and code reasoning tasks.
Large Language Model Transformers
M
XiaomiMiMo
1,183
23
Tngtech.olmo 2 Instruct Math 32B GGUF
OLMo-2-Instruct-Math-32B is a large language model focused on mathematical tasks, released by tngtech.
Large Language Model
T
DevQuasar
272
1
Openmath Nemotron 1.5B
OpenMath-Nemotron-1.5B is a mathematical reasoning model fine-tuned on the OpenMathReasoning dataset based on Qwen2.5-Math-1.5B, achieving state-of-the-art results on multiple mathematical benchmarks.
Large Language Model Transformers English
O
nvidia
493
14
Zero Mistral 24B
MIT
Zero-Mistral-24B is an improved text-only model based on Mistral-Small-3.1-24B-Instruct-2503, primarily adapted for Russian and English, with the original visual capabilities removed to focus on text generation tasks.
Large Language Model Transformers Supports Multiple Languages
Z
ZeroAgency
41
2
Openmath2 Llama3.1 8B
OpenMath2-Llama3.1-8B is a math-specialized model fine-tuned from the Llama3.1-8B-Base model using the OpenMathInstruct-2 dataset, demonstrating excellent performance across multiple mathematical benchmarks.
Large Language Model Transformers English
O
nvidia
930
30
Viper Coder V1.7 Vsm6
Apache-2.0
Viper-Coder-v1.7-Vsm6 is a large language model based on the Qwen2.5 14B modal architecture, focusing on improving coding efficiency and computational reasoning capabilities, optimizing memory usage, and reducing redundant text generation.
Large Language Model Transformers Supports Multiple Languages
V
prithivMLmods
491
5
Phi 4 Reasoning Plus
MIT
Phi-4-reasoning-plus is an advanced open-weight reasoning model developed by Microsoft Research, optimized through supervised fine-tuning and reinforcement learning based on Phi-4, focusing on advanced reasoning capabilities in mathematics, science, and coding fields.
Large Language Model Transformers Supports Multiple Languages
P
microsoft
19.83k
261
EZO2.5 Gemma 3 12b It Preview
A text generation model based on google/gemma-3-12b-it, enhanced with the EZO training method to improve Japanese performance
Large Language Model Transformers Japanese
E
AXCXEPT
39
1
Notbad V1 1 Mistral 24b
Apache-2.0
A 24B-parameter large language model optimized for mathematical reasoning and Python programming training, based on the Mistral architecture
Large Language Model Transformers
N
notbadai
34
4
Openrs3 GRPO Ja
OpenRS3-GRPO-ja is a fine-tuned version of the SakanaAI/TinySwallow-1.5B-Instruct model on a Japanese mathematical instruction dataset, trained using the GRPO method, focusing on mathematical reasoning tasks.
Large Language Model Transformers
O
EQUES
25
3
Notbad V1 0 Mistral 24b
Apache-2.0
Notbad v1.0 Mistral 24B is a model focused on mathematical and Python programming reasoning, based on Mistral-Small-24B-Instruct-2501 and further trained with reinforcement learning.
Large Language Model Transformers
N
notbadai
29
5
EXAONE Deep 7.8B GGUF
Other
The EXAONE Deep series models excel in reasoning tasks such as mathematics and programming. The 7.8B version outperforms open-source models of similar scale and even surpasses certain proprietary models.
Large Language Model Supports Multiple Languages
E
QuantFactory
297
3
Olmo 2 Instruct Math 32B
Apache-2.0
Based on the OLMo-2-0325-32B-Instruct model, fine-tuned using the Open R1 math dataset on AMD MI300X GPUs, focusing on enhancing mathematical reasoning capabilities
Large Language Model Transformers English
O
tngtech
96
5
Fastcurl 1.5B Preview
MIT
FastCuRL-1.5B-Preview is a slow-thinking reasoning model that employs curriculum-guided iterative extended reinforcement learning, excelling in mathematical reasoning tasks.
Large Language Model Transformers English
F
Nickyang
779
7
Yixin Distill Qwen 72B 4.5bpw H6 Exl2
Apache-2.0
A high-performance mathematical reasoning and general knowledge processing model distilled from Qwen2.5-72B through reinforcement learning, excelling in mathematical reasoning and general knowledge tasks.
Large Language Model Supports Multiple Languages
Y
LoneStriker
37
3
Gemma 3 4b Reasoning
Apache-2.0
Gemma-3-4b Reasoning is a Transformer-based language model fine-tuned using the GRPO method, specializing in reasoning task optimization.
Large Language Model Transformers English
G
ericrisco
53
2
Yixin Distill Qwen 72B
Apache-2.0
A high-performance distilled model optimized for mathematics and general reasoning, refined from Qwen2.5-72B through reinforcement learning
Large Language Model Safetensors Supports Multiple Languages
Y
YiXin-AILab
38
26
Qwen 2.5 7B Reasoning
MIT
A fine-tuned version based on Qwen/Qwen2.5-7B-Instruct, specifically optimized for advanced reasoning tasks
Large Language Model Transformers English
Q
HyperX-Sen
70
3
Sombrero Opus 14B Sm5
Apache-2.0
Designed based on Qwen 2.5 14B modal architecture, enhancing coding efficiency and computational reasoning capabilities
Large Language Model Transformers Supports Multiple Languages
S
prithivMLmods
43
2
Tinyr1 32B Preview
Apache-2.0
Tiny-R1-32B-Preview is an inference model based on Deepseek-R1-Distill-Qwen-32B, focusing on mathematics, coding, and scientific fields, with performance close to the full version of the R1 model.
Large Language Model Transformers
T
qihoo360
3,292
327
Mistral Small 24B Instruct 2501 Reasoning
Apache-2.0
A mathematical reasoning model fine-tuned based on Mistral-Small-24B-Instruct-2501, optimized for mathematical reasoning capabilities
Large Language Model Safetensors English
M
yentinglin
1,689
54
Sky T1 32B Flash
Apache-2.0
An optimized 32B inference model preference version based on Sky-T1-32B-Preview, significantly reducing generation length while maintaining accuracy.
Large Language Model Transformers English
S
NovaSky-AI
557
64
Internlm3 8b Instruct Gguf
Apache-2.0
The GGUF format version of the InternLM3-8B-Instruct model, suitable for the llama.cpp framework and supporting multiple quantization versions.
Large Language Model English
I
internlm
1,072
26
Tulu3
Other
Tülu3 is a new generation of instruction-following model family developed by the Allen Institute for Artificial Intelligence, excelling in standard chat applications and complex problem-solving.
Large Language Model
T
cortexso
226
1
Rho Math 1b V0.1
MIT
Rho-1 is a language model specialized in mathematics, pretrained with Selective Language Modeling (SLM) method, significantly improving accuracy in solving mathematical problems.
Large Language Model Transformers English
R
microsoft
1,451
15
UNA SimpleSmaug 34b V1beta
Apache-2.0
A supervised fine-tuned model based on Smaug-34B, focused on enhancing mathematical and reasoning capabilities, excelling among 34B-scale models.
Large Language Model Transformers
U
fblgit
18
21
Westseverus 7B DPO V2
Apache-2.0
WestSeverus-7B-DPO-v2 is a WestLake family model trained based on WestSeverus-7B, with training on multiple DPO datasets, demonstrating excellent performance in basic mathematical problems.
Large Language Model Transformers English
W
PetroGPT
437
8
Neural Chat 7b V3 3
Apache-2.0
Neural-Chat-v3-3 is a 7-billion-parameter large language model developed by Intel based on the Mistral-7B architecture, focusing on mathematical reasoning and text generation tasks. The model is fine-tuned on the MetaMathQA dataset and aligned using Direct Performance Optimization (DPO) method.
Large Language Model Transformers
N
Intel
29.82k
78
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase